Dynamic Regret of Online Mirror Descent for Relatively Smooth Convex Cost Functions

نویسندگان

چکیده

The performance of online convex optimization algorithms in a dynamic environment is often expressed terms the regret, which measures decision maker’s against sequence time-varying comparators. In analysis prior works assume Lipschitz continuity or uniform smoothness cost functions. However, there are many important functions practice that do not satisfy these conditions. such cases, analyses applicable and fail to guarantee performance. this letter, we show it possible bound even when neither nor present. We adopt notion relative with respect some user-defined regularization function, much milder requirement on first under smoothness, regret has an upper based path length functional variation. then additional condition relatively strong convexity, can be bounded by gradient These bounds provide guarantees wide variety problems arise different application domains. Finally, present numerical experiments demonstrate advantage adopting function smooth.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Shifting Regret, Mirror Descent, and Matrices

We consider the problem of online prediction in changing environments. In this framework the performance of a predictor is evaluated as the loss relative to an arbitrarily changing predictor, whose individual components come from a base class of predictors. Typical results in the literature consider different base classes (experts, linear predictors on the simplex, etc.) separately. Introducing...

متن کامل

No-regret Algorithms for Online Convex Programs

Online convex programming has recently emerged as a powerful primitive for designing machine learning algorithms. For example, OCP can be used for learning a linear classifier, dynamically rebalancing a binary search tree, finding the shortest path in a graph with unknown edge lengths, solving a structured classification problem, or finding a good strategy in an extensive-form game. Several res...

متن کامل

No-regret algorithms for Online Convex Programs

Online convex programming has recently emerged as a powerful primitive for designing machine learning algorithms. For example, OCP can be used for learning a linear classifier, dynamically rebalancing a binary search tree, finding the shortest path in a graph with unknown edge lengths, solving a structured classification problem, or finding a good strategy in an extensive-form game. Several res...

متن کامل

Mirror descent in non-convex stochastic programming

In this paper, we examine a class of nonconvex stochastic optimization problems which we call variationally coherent, and which properly includes all quasi-convex programs. In view of solving such problems, we focus on the widely used stochastic mirror descent (SMD) family of algorithms, and we establish that the method’s last iterate converges with probability 1. We further introduce a localiz...

متن کامل

Convergence of Online Mirror Descent Algorithms

In this paper we consider online mirror descent (OMD) algorithms, a class of scalable online learning algorithms exploiting data geometric structures through mirror maps. Necessary and sufficient conditions are presented in terms of the step size sequence {ηt}t for the convergence of an OMD algorithm with respect to the expected Bregman distance induced by the mirror map. The condition is limt→...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Control Systems Letters

سال: 2022

ISSN: ['2475-1456']

DOI: https://doi.org/10.1109/lcsys.2022.3155067